573 research outputs found

    Person Re-identification with Correspondence Structure Learning

    Full text link
    This paper addresses the problem of handling spatial misalignments due to camera-view changes or human-pose variations in person re-identification. We first introduce a boosting-based approach to learn a correspondence structure which indicates the patch-wise matching probabilities between images from a target camera pair. The learned correspondence structure can not only capture the spatial correspondence pattern between cameras but also handle the viewpoint or human-pose variation in individual images. We further introduce a global-based matching process. It integrates a global matching constraint over the learned correspondence structure to exclude cross-view misalignments during the image patch matching process, hence achieving a more reliable matching score between images. Experimental results on various datasets demonstrate the effectiveness of our approach

    Normal families and shared values of meromorphic functions

    Get PDF
    AbstractLet k(⩾2) be a positive integer, let F be a family of meromorphic functions in a domain D, all of whose zeros have multiplicity at least k+1, and let a(z)(≠0), h(z)(≢0) be two holomorphic functions on D. If, for each f∈F, f=a(z)⇔f(k)=h(z), then F is normal in D

    Learning Correspondence Structures for Person Re-identification

    Full text link
    This paper addresses the problem of handling spatial misalignments due to camera-view changes or human-pose variations in person re-identification. We first introduce a boosting-based approach to learn a correspondence structure which indicates the patch-wise matching probabilities between images from a target camera pair. The learned correspondence structure can not only capture the spatial correspondence pattern between cameras but also handle the viewpoint or human-pose variation in individual images. We further introduce a global constraint-based matching process. It integrates a global matching constraint over the learned correspondence structure to exclude cross-view misalignments during the image patch matching process, hence achieving a more reliable matching score between images. Finally, we also extend our approach by introducing a multi-structure scheme, which learns a set of local correspondence structures to capture the spatial correspondence sub-patterns between a camera pair, so as to handle the spatial misalignments between individual images in a more precise way. Experimental results on various datasets demonstrate the effectiveness of our approach.Comment: IEEE Trans. Image Processing, vol. 26, no. 5, pp. 2438-2453, 2017. The project page for this paper is available at http://min.sjtu.edu.cn/lwydemo/personReID.htm arXiv admin note: text overlap with arXiv:1504.0624

    Changes in respiratory structure and function after traumatic cervical spinal cord injury: observations from spinal cord and brain

    Get PDF
    Respiratory difficulties and mortality following severe cervical spinal cord injury (CSCI) result primarily from malfunctions of respiratory pathways and the paralyzed diaphragm. Nonetheless, individuals with CSCI can experience partial recovery of respiratory function through respiratory neuroplasticity. For decades, researchers have revealed the potential mechanism of respiratory nerve plasticity after CSCI, and have made progress in tissue healing and functional recovery. While most existing studies on respiratory plasticity after spinal cord injuries have focused on the cervical spinal cord, there is a paucity of research on respiratory-related brain structures following such injuries. Given the interconnectedness of the spinal cord and the brain, traumatic changes to the former can also impact the latter. Consequently, are there other potential therapeutic targets to consider? This review introduces the anatomy and physiology of typical respiratory centers, explores alterations in respiratory function following spinal cord injuries, and delves into the structural foundations of modified respiratory function in patients with CSCI. Additionally, we propose that magnetic resonance neuroimaging holds promise in the study of respiratory function post-CSCI. By studying respiratory plasticity in the brain and spinal cord after CSCI, we hope to guide future clinical work

    Self-supervised Point Cloud Representation Learning via Separating Mixed Shapes

    Full text link
    The manual annotation for large-scale point clouds costs a lot of time and is usually unavailable in harsh real-world scenarios. Inspired by the great success of the pre-training and fine-tuning paradigm in both vision and language tasks, we argue that pre-training is one potential solution for obtaining a scalable model to 3D point cloud downstream tasks as well. In this paper, we, therefore, explore a new self-supervised learning method, called Mixing and Disentangling (MD), for 3D point cloud representation learning. As the name implies, we mix two input shapes and demand the model learning to separate the inputs from the mixed shape. We leverage this reconstruction task as the pretext optimization objective for self-supervised learning. There are two primary advantages: 1) Compared to prevailing image datasets, eg, ImageNet, point cloud datasets are de facto small. The mixing process can provide a much larger online training sample pool. 2) On the other hand, the disentangling process motivates the model to mine the geometric prior knowledge, eg, key points. To verify the effectiveness of the proposed pretext task, we build one baseline network, which is composed of one encoder and one decoder. During pre-training, we mix two original shapes and obtain the geometry-aware embedding from the encoder, then an instance-adaptive decoder is applied to recover the original shapes from the embedding. Albeit simple, the pre-trained encoder can capture the key points of an unseen point cloud and surpasses the encoder trained from scratch on downstream tasks. The proposed method has improved the empirical performance on both ModelNet-40 and ShapeNet-Part datasets in terms of point cloud classification and segmentation tasks. We further conduct ablation studies to explore the effect of each component and verify the generalization of our proposed strategy by harnessing different backbones
    corecore